skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Search for: All records

Creators/Authors contains: "Xu, Xu"

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. Generative AI holds promise for advancing human factors and ergonomics. However, limited training data can reduce variability in generative models. This study investigates using transfer learning to enhance variability in generative posture prediction. We pre-trained a conditional diffusion model on lifting postures where hands are near body center and fine-tuned it on limited extended-reach postures. Compared to training from scratch, transfer learning significantly improved joint angle variability across multiple body segments while maintaining similar accuracy in posture similarity and validity. Additionally, it reduced training time by over 90%, demonstrating efficiency benefits. These findings highlight transfer learning’s potential to enrich generative model outputs with more variable ergonomics data, supporting scalable and adaptive workplace design. 
    more » « less
  2. Workers performing repetitive lifting tasks are at high risk of developing low-back work-related musculoskeletal disorders. While the Revised NIOSH Lifting Equation (RNLE) is a widely used tool for evaluating lifting-related risks, its reliance on manual measurement limits its scalability and efficiency. This study proposes a computer vision-based framework that automates RNLE computation using video data. The method integrates three key stages: (1) pose estimation to extract 3D joint coordinates, (2) lifting action recognition via kinematic features and a k-TSP classifier, and (3) estimation of RNLE multipliers from joint data. Applied to 40 lifting trials with motion capture-based ground-truth, the system achieved a coefficient of determination of 0.82 and a mean absolute error of 2.72 kg in estimating recommended weight limits. These findings demonstrate the potential of computer vision to automate ergonomic risk assessments. Future work will aim to expand task diversity and integrate coupling assessment for full RNLE coverage. 
    more » « less
  3. Free, publicly-accessible full text available August 1, 2026
  4. ObjectiveTo identify lifting actions and count the number of lifts performed in videos based on robust class prediction and a streamlined process for reliable real-time monitoring of lifting tasks. BackgroundTraditional methods for recognizing lifting actions often rely on deep learning classifiers applied to human motion data collected from wearable sensors. Despite their high performance, these methods can be difficult to implement on systems with limited hardware resources. MethodThe proposed method follows a five-stage process: (1) BlazePose, a real-time pose estimation model, detects key joints of the human body. (2) These joints are preprocessed by smoothing, centering, and scaling techniques. (3) Kinematic features are extracted from the preprocessed joints. (4) Video frames are classified as lifting or nonlifting using rank-altered kinematic feature pairs. (5) A lifting counting algorithm counts the number of lifts based on the class predictions. ResultsNine rank-altered kinematic feature pairs are identified as key pairs. These pairs were used to construct an ensemble classifier, which achieved 0.89 or above in classification metrics, including accuracy, precision, recall, and F1 score. This classifier showed an accuracy of 0.90 in lifting counting and a latency of 0.06 ms, which is at least 12.5 times faster than baseline classifiers. ConclusionThis study demonstrates that computer vision-based kinematic features could be adopted to effectively and efficiently recognize lifting actions. ApplicationThe proposed method could be deployed on various platforms, including mobile devices and embedded systems, to monitor lifting tasks in real-time for the proactive prevention of work-related low-back injuries. 
    more » « less
  5. This study presents a mobile app that facilitates undergraduate students to learn data science through their own full body motions. Leveraging the built-in camera of a mobile device, the proposed app captures the user and feeds their images into an open-source computer-vision algorithm that localizes the key joint points of human body. As students can participate in the entire data collection process, the obtained motion data is context-rich and personally relevant to them. The app utilizes the collected motion data to explain various concepts and methods in data science under the context of human movements. The app also visualizes the geometric interpretation of data through various visual aids, such as interactive graphs and figures. In this study, we use principal component analysis, a commonly used dimensionality reduction method, as an example to demonstrate the proposed learning framework. Strategies to encompass other learning modules are also discussed for further improvement. 
    more » « less
  6. In recent years, there has been a trend to adopt human-robot collaboration (HRC) in the industry. In previous studies, computer vision-aided human pose reconstruction is applied to find the optimal position of point of operation in HRC that can reduce workers’ musculoskeletal disorder (MSD) risks due to awkward working postures. However, the reconstruction of human pose through computer-vision may fail due to the complexity of the workplace environment. In this study, we propose a data-driven method for optimizing the position of point of operation during HRC. A conditional variational auto-encoder (cVAE) model-based approach is adopted, which includes three steps. First, a cVAE model was trained using an open-access multimodal human posture dataset. After training, this model can output a simulated worker posture of which the hand position can reach a given position of point of operation. Next, an awkward posture score is calculated to evaluate MSD risks associated with the generated postures with a variety of positions of point of operation. The position of point of operation that is associated with a minimum awkward posture score is then selected for an HRC task. An experiment was conducted to validate the effectiveness of this method. According to the findings, the proposed method produced a point of operation position that was similar to the one chosen by participants through subjective selection, with an average difference of 4.5 cm. 
    more » « less